# GPU-efficient inference
BAGEL 7B MoT DF11
The BAGEL-7B-MoT model utilizing DFloat11 lossless compression technology reduces volume by 32% while maintaining bit-level output consistency
Large Language Model
Safetensors
B
DFloat11
428
8
Paraformer Large
Apache-2.0
Paraformer is an innovative non-autoregressive end-to-end speech recognition model with significant advantages over traditional autoregressive models. It can generate entire target sentences in parallel, making it particularly suitable for GPU-accelerated parallel inference.
Speech Recognition Chinese
P
funasr
43
45
Gpt J 6B
GPT-J 6B is a Transformer model based on the GPT-3 architecture, featuring 6 billion parameters and supporting text generation tasks.
Large Language Model
Transformers

G
flyhero
59
13
Featured Recommended AI Models